skip to main content


Search for: All records

Creators/Authors contains: "Stein, George"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. ABSTRACT

    Gravitational lensing deflects the paths of photons, altering the statistics of cosmic backgrounds and distorting their information content. We take the cosmic infrared background (CIB), which provides plentiful information about galaxy formation and evolution, as an example to probe the effect of lensing on non-Gaussian statistics. Using the Websky simulations, we first quantify the non-Gaussianity of the CIB, revealing additional detail on top of its well-measured power spectrum. To achieve this, we use needlet-like multipole-band filters to calculate the variance and higher-point correlations. Using our simulations, we show the two-, three- and four-point spectra, and compare our calculated power spectra and bispectra to Planck values. We then lens the CIB, shell-by-shell with corresponding convergence maps, to capture the broad redshift extent of both the CIB and its lensing convergence. The lensing of the CIB changes the three- and four-point functions by a few tens of per cent at large scales, unlike with the power spectrum, which changes by less than two per cent. We expand our analyses to encompass the full intensity probability distribution functions (PDFs) involving all n-point correlations as a function of scale. In particular, we use the relative entropy between lensed and unlensed PDFs to create a spectrum of templates that can allow estimation of lensing. The underlying CIB model is missing the important role of star bursting, which we test by adding a stochastic lognormal term to the intensity distributions. The novel aspects of our filtering and lensing pipeline should prove useful for any radiant background, including line intensity maps.

     
    more » « less
  2. Abstract We construct a physically parameterized probabilistic autoencoder (PAE) to learn the intrinsic diversity of Type Ia supernovae (SNe Ia) from a sparse set of spectral time series. The PAE is a two-stage generative model, composed of an autoencoder that is interpreted probabilistically after training using a normalizing flow. We demonstrate that the PAE learns a low-dimensional latent space that captures the nonlinear range of features that exists within the population and can accurately model the spectral evolution of SNe Ia across the full range of wavelength and observation times directly from the data. By introducing a correlation penalty term and multistage training setup alongside our physically parameterized network, we show that intrinsic and extrinsic modes of variability can be separated during training, removing the need for the additional models to perform magnitude standardization. We then use our PAE in a number of downstream tasks on SNe Ia for increasingly precise cosmological analyses, including the automatic detection of SN outliers, the generation of samples consistent with the data distribution, and solving the inverse problem in the presence of noisy and incomplete data to constrain cosmological distance measurements. We find that the optimal number of intrinsic model parameters appears to be three, in line with previous studies, and show that we can standardize our test sample of SNe Ia with an rms of 0.091 ± 0.010 mag, which corresponds to 0.074 ± 0.010 mag if peculiar velocity contributions are removed. Trained models and codes are released at https://github.com/georgestein/suPAErnova. 
    more » « less
  3. Kasieczka, Gregor ; Nachman, Benjamin ; Shih, David (Ed.)
    A new paradigm for data-driven, model-agnostic new physics searches at colliders is emerging, and aims to leverage recent breakthroughs in anomaly detection and machine learning. In order to develop and benchmark new anomaly detection methods within this framework, it is essential to have standard datasets. To this end, we have created the LHC Olympics 2020, a community challenge accompanied by a set of simulated collider events. Participants in these Olympics have developed their methods using an R&D dataset and then tested them on black boxes: datasets with an unknown anomaly (or not). This paper will review the LHC Olympics 2020 challenge, including an overview of the competition, a description of methods deployed in the competition, lessons learned from the experience, and implications for data analyses with future datasets as well as future colliders. 
    more » « less